Linearly Replaceable Filters for Deep Network Channel Pruning
نویسندگان
چکیده
Convolutional neural networks (CNNs) have achieved remarkable results; however, despite the development of deep learning, practical user applications are fairly limited because heavy can be used solely with latest hardware and software supports. Therefore, network pruning is gaining attention for general in various fields. This paper proposes a novel channel method, Linearly Replaceable Filter (LRF), which suggests that filter approximated by linear combination other filters replaceable. Moreover, an additional method called Weights Compensation proposed to support LRF method. technique effectively reduces output difference caused removing via direct weight modification. Through experiments, we confirmed our achieves state-of-the-art performance several benchmarks. In particular, on ImageNet, LRF-60 approximately 56% FLOPs ResNet-50 without top-5 accuracy drop. Further, through extensive analyses, proved effectiveness approaches.
منابع مشابه
Automated Pruning for Deep Neural Network Compression
In this work we present a method to improve the pruning step of the current state-of-the-art methodology to compress neural networks. The novelty of the proposed pruning technique is in its differentiability, which allows pruning to be performed during the backpropagation phase of the network training. This enables an end-to-end learning and strongly reduces the training time. The technique is ...
متن کاملPruning Filters for Efficient ConvNets
Convolutional Neural Networks (CNNs) are extensively used in image and video recognition, natural language processing and other machine learning applications. The success of CNNs in these areas corresponds with a significant increase in the number of parameters and computation costs. Recent approaches towards reducing these overheads involve pruning and compressing the weights of various layers...
متن کاملAutomatic Tuning for Linearly Tunable Filters
Automatic Tuning for Linearly Tunable Filters. (May 2004) Sung-Ling Huang, B.S., Tatung Institute of Technology, Taiwan; M.S., Tatung Institute of Technology, Taiwan Chair of Advisory Committee: Dr. Aydin I. Karsilayan A new tuning scheme for linearly tunable high-Q filters is proposed. The tuning method is based on using the phase information for both frequency and Q factor tuning. There is no...
متن کاملStructured Deep Neural Network Pruning via Matrix Pivoting
Deep Neural Networks (DNNs) are the key to the state-of-the-art machine vision, sensor fusion and audio/video signal processing. Unfortunately, their computation complexity and tight resource constraints on the Edge make them hard to leverage on mobile, embedded and IoT devices. Due to great diversity of Edge devices, DNN designers have to take into account the hardware platform and application...
متن کاملA Deep Neural Network Compression Pipeline: Pruning, Quantization, Huffman Encoding
Neural networks are both computationally intensive and memory intensive, making them difficult to deploy on embedded systems with limited hardware resources. To address this limitation, We introduce a three stage pipeline: pruning, quantization and Huffman encoding, that work together to reduce the storage requirement of neural networks by 35× to 49× without affecting their accuracy. Our method...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2021
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v35i9.16978